Optimization of High Order Perceptrons
نویسندگان
چکیده
Neural networks are widely applied in research and industry However their broader ap plication is hampered by various technical details Among these details are several training parameters and the choice of the topology of the network The subject of this dissertation is therefore the elimination and determination of usually user speci ed learning parameters Furthermore suitable application domains for neural networks are discussed Among all training parameters special attention is given to the learning rate the gain of the sigmoidal function and the initial weight range A theorem is proven which permits the elimination of one of these parameters Furthermore it is shown that for high order perceptrons very small random initial weights are usually optimal in terms of training time and generalization Another important problem in the application of neural networks is to nd a network topology that suits a given data set This favors high order perceptrons over several other neural network architectures as they do not require layers of hidden neurons However the order and the connectivity of a network have to be determined which is possible by two approaches The rst is to remove connections from an initially big network while training it The other approach is to increase gradually the network size Both types of approaches are studied corresponding algorithms are developed and applied to high order perceptrons The dis advantages of both approaches are gone into and their performance experimentally compared Then an outlook on future research on the interpretation and analysis of high order perceptrons and their feasibility is given Finally high order perceptrons and the developed algorithms are applied to a number of real world applications and in order to show their e ciency the obtained performances are compared to those of other approaches
منابع مشابه
Bounds on the Degree of High Order Binary Perceptrons Bounds on the Degree of High Order Binary Perceptrons
High order perceptrons are often used in order to reduce the size of neural networks. The complexity of the architecture of a usual multilayer network is then turned into the complexity of the functions performed by each high order unit and in particular by the degree of their polynomials. The main result of this paper provides a bound on the degree of the polynomial of a high order perceptron,...
متن کاملBounds on the degree of high order binary perceptrons
High order perceptrons are often used in order to reduce the size of neural networks The complexity of the architecture of a usual multilayer network is then turned into the complexity of the functions performed by each high order unit and in particular by the degree of their polynomials The main result of this paper provides a bound on the degree of the polynomial of a high order perceptron wh...
متن کاملHyperparameter Optimization with Factorized Multilayer Perceptrons
In machine learning, hyperparameter optimization is a challenging task that is usually approached by experienced practitioners or in a computationally expensive brute-force manner such as grid-search. Therefore, recent research proposes to use observed hyperparameter performance on already solved problems (i.e. data sets) in order to speed up the search for promising hyperparameter configuratio...
متن کاملAdaptive High Order Neural Trees for Pattern Recognition
In this paper, a new classifier, called adaptive high order neural tree (AHNT), is proposed for pattern recognition applications. It is a hierarchical multi-level neural network, in which the nodes are organized into a tree topology. It successively partitions the training set into subsets, assigning each subset to a different child node. Each node can be a first-order or a high order perceptro...
متن کاملThe Design and Complexity of Exact Multilayered Perceptrons
We investigate the network complexity of multi-layered perceptrons for solving exactly a given problem. We limit our study to the class of combinatorial optimization problems. It is shown how these problems can be reformulated as binary classification problems and how they can be solved by multi-layered perceptrons.
متن کامل